Local-aware spatio-temporal attention network with multi-stage feature fusion for human action recognition

نویسندگان

چکیده

Abstract In the study of human action recognition, two-stream networks have made excellent progress recently. However, there remain challenges in distinguishing similar actions videos. This paper proposes a novel local-aware spatio-temporal attention network with multi-stage feature fusion based on compact bilinear pooling for recognition. To elaborate, taking as our essential backbones, spatial first employs multiple transformer parallel manner to locate discriminative regions related actions. Then, we perform between local and global features enhance representation. Furthermore, output temporal information are fused at particular layer learn pixel-wise correspondences. After that, bring together three outputs generate descriptors verify efficacy proposed approach, comparison experiments conducted traditional hand-engineered IDT algorithms, classical machine learning methods (i.e., SVM) state-of-the-art deep multiplier networks). According results, approach is reported obtain best performance among existing works, accuracy 95.3% 72.9% UCF101 HMDB51, respectively. The experimental results thus demonstrate superiority significance architecture solving task

برای دانلود باید عضویت طلایی داشته باشید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Pose-conditioned Spatio-Temporal Attention for Human Action Recognition

We address human action recognition from multi-modal video data involving articulated pose and RGB frames and propose a two-stream approach. The pose stream is processed with a convolutional model taking as input a 3D tensor holding data from a sub-sequence. A specific joint ordering, which respects the topology of the human body, ensures that different convolutional layers correspond to meanin...

متن کامل

Human action categorization using discriminative local spatio-temporal feature weighting

New methods based on local spatio-temporal features have exhibited significant performance in action recognition. In these methods, feature selection plays an important role to achieve a superior performance. Actions are represented by local spatio-temporal features extracted from action videos. Action representations are then classified by applying a classifier (such as k-nearest neighbor or S...

متن کامل

A SURF-Based Spatio-Temporal Feature for Feature-Fusion-Based Action Recognition

In this paper, we propose a novel spatio-temporal feature which is useful for feature-fusion-based action recognition with Multiple Kernel Learning (MKL). The proposed spatio-temporal feature is based on moving SURF interest points grouped by Delaunay triangulation and on their motion over time. Since this local spatio-temporal feature has different characteristics from holistic appearance feat...

متن کامل

Spatio-temporal SURF for Human Action Recognition

In this paper, we propose a new spatio-temporal descriptor called ST-SURF. The latter is based on a novel combination between the speed up robust feature and the optical flow. The Hessian detector is employed to find all interest points. To reduce the computation time, we propose a new methodology for video segmentation, in Frames Packets FPs, based on the interest points trajectory tracking. W...

متن کامل

Spatio-temporal Aware Non-negative Component Representation for Action Recognition

This paper presents a novel mid-level representation for action recognition, named spatio-temporal aware non-negative component representation (STANNCR). The proposed STANNCR is based on action component and incorporates the spatial-temporal information. We first introduce a spatial-temporal distribution vector (STDV) to model the distributions of local feature locations in a compact and discri...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Neural Computing and Applications

سال: 2021

ISSN: ['0941-0643', '1433-3058']

DOI: https://doi.org/10.1007/s00521-021-06239-5